# Multi-query attention
Bert 1.3b
Apache-2.0
Transformer encoder pretrained based on Megatron-LM, specifically designed for Japanese scenarios
Large Language Model
Transformers Supports Multiple Languages

B
retrieva-jp
56
15
Falcon 7B Instruct GPTQ
Apache-2.0
The 4-bit quantized version of Falcon-7B-Instruct, quantized using the AutoGPTQ tool, suitable for efficient inference in resource-constrained environments.
Large Language Model
Transformers English

F
TheBloke
189
67
Featured Recommended AI Models